Time-of-flight camera

A time-of-flight camera (ToF camera) is a range imaging camera system that resolves distance based on the known speed of light, measuring the time-of-flight of a light signal between the camera and the subject for each point of the image. The time-of-flight camera is a class of scannerless LIDAR, in which the entire scene is captured with each laser or light pulse, as opposed to point-by-point with a laser beam such as in scanning LIDAR systems.[1]

Time-of-flight camera products for civil applications began to emerge around 2000,[2] as the semiconductor processes became fast enough for such devices. The systems cover ranges of a few meters up to about several kilometers depending upon the detector material being used.

The imager module is made from (1) CMOS detectors which capture light pulses in the visible range, (2) PIN diodes or (3) avalanche photo diodes (APDs). Each material has inherent strengths and is chosen based upon application requirements.

The distance resolution ranges from sub-centimeter to several centimeters depending upon the range. The lateral resolution of time-of-flight cameras is generally low compared to standard 2D video cameras, with most commercially available devices at 320 × 240 pixels or less as of 2011.[3][4][5][6][7] Compared to 3D laser scanning methods for capturing 3D images, TOF cameras operate very quickly, providing up to 100 images per second.[8]

Contents

Types of devices

Several different technologies for time-of-flight cameras have been developed.

Pulsed light source with digital time counters

There are devices with a pulsed laser and a custom imaging integrated circuit with a fast counter behind every pixel. These devices produce depth values for each pixel on every frame. Typical image sizes are 128 x 128 pixels. Ranges up to 22,000 feet with an eye-safe narrow beam have been achieved. Detectors are typically InGaAs (indium-gallium-arsenide) devices.[9]

RF-modulated light sources with phase detectors

Photonic Mixer Devices (PMD),[10] the Swiss Ranger, and CanestaVision[11] work by modulating the outgoing beam with an RF carrier, then measuring the phase shift of that carrier on the receive side. This approach has a modular error challenge; ranges are mod the maximum range, which is the RF carrier wavelength. The Swiss Ranger is a compact, short-range device, with ranges of 5 or 10 meters, with 176 x 144 pixels. With phase unwrapping algorithms, the maximum uniqueness range can be increased. The PMD can provide ranges up to 60m. Illumination is pulsed LEDs, rather than a laser.[12] CanestaVision developer Canesta was purchased by Microsoft in 2010.

Range gated imagers

These devices have a built-in shutter in front of the image sensor that opens and closes at the same rate as the light pulses are sent out. Because part of every returning pulse is blocked by the shutter according to its time of arrival, the amount of light received relates to the distance the pulse has traveled. The distance can be calculated using the equation, z = R (S2S1) / 2(S1 + S2) + R / 2 for an ideal camera. R is the camera range, determined by the round trip of the light pulse, S1 the amount of the light pulse that is received, and S2 the amount of the light pulse that is blocked.[13][14]

The ZCam by 3DV Systems[1] is a range-gated system. Microsoft purchased 3DV in 2009.

Similar principles are used in the ToF camera line developed by the Fraunhofer Institute of Microelectronic Circuits and Systems and TriDiCam. These cameras employ photodetectors with a fast electronic shutter.

Range gated imagers can also be used in 2D imaging to suppress anything outside a specified distance range, such as to see through fog. A pulsed laser provides illumination, and an optical gate allows light to reach the imager only during the desired time period.[15][16]

Components

A time-of-flight camera consists of the following components:

Principle

The simplest version of a time-of-flight camera uses light pulses. The illumination is switched on for a very short time, the resulting light pulse illuminates the scene and is reflected by the objects. The camera lens gathers the reflected light and images it onto the sensor plane. Depending on the distance, the incoming light experiences a delay. As light has a speed of approximately c = 300,000,000 meters per second, this delay is very short: an object 2.5 m away will delay the light by:

t_D = 2 \cdot \frac D c = 2 \cdot \frac {2.5\;\mathrm{m}} {300\;000\;000\;\frac{\mathrm{m}}{\mathrm{s}}} = 0.000\;000\;016\;66\;\mathrm{s} = 16.66 \;\mathrm{ns}[17]

The pulse width of the illumination determines the maximum range the camera can handle. With a pulse width of e.g. 50 ns, the range is limited to

D_\mathrm{max} = \frac{1}{2} \cdot c \cdot t_0 = \frac{1}{2} \cdot 300\;000\;000\;\frac{\mathrm{m}}{\mathrm{s}} \cdot 0.000\;000\;05\;\mathrm{s} =\!\ 7.5\;\mathrm{m}

These short times show that the illumination unit is a critical part of the system. Only with some special LEDs or lasers is it possible to generate such short pulses.

The single pixel consists of a photo sensitive element (e.g. a photo diode). It converts the incoming light into a current. In analog timing imagers, connected to the photo diode are fast switches, which direct the current to one of two (or several) memory elements (e.g. a capacitor) that act as summation elements. In digital timing imagers, a time counter, running at several gigahertz, is connected to each photodetector pixel and stops counting when light is sensed.

In the diagram of an analog timer, the pixel uses two switches (G1 and G2) and two memory elements (S1 and S2). The switches are controlled by a pulse with the same length as the light pulse, where the control signal of switch G2 is delayed by exactly the pulse width. Depending on the delay, only part of the light pulse is sampled through G1 in S1, the other part is stored in S2. Depending on the distance, the ratio between S1 and S2 changes as depicted in the drawing.[11] Because only small amounts of light hit the sensor within 50 ns, not only one but several thousands pulses are sent out (repetition rate tR) and gathered, thus increasing the signal to noise ratio.

After the exposure, the pixel is read out and the following stages measure the signals S1 and S2. As the length of the light pulse is defined, the distance can be calculated with the formula:

D = \frac{1}{2} \cdot c \cdot t_0 \cdot \frac {S2} {S1 %2B S2}

In the example, the signals have the following values: S1 = 0.66 and S2 = 0.33. The distance is therefore:

D = 7.5\;\mathrm{m} \cdot \frac {0.33} {0.33 %2B 0.66} = 2.5\;\mathrm{m}

In the presence of background light, the memory elements receive an additional part of the signal. This would disturb the distance measurement. To eliminate the background part of the signal, the whole measurement can be performed a second time with the illumination switched off. If the objects are further away than the distance range, the result is also wrong. Here, a second measurement with the control signals delayed by an additional pulse width helps to suppress such objects. Other systems work with a sinusoidally modulated light source instead of the pulse source.

Advantages

Simplicity

In contrast to stereo vision or triangulation systems, the whole system is very compact: the illumination is placed just next to the lens, whereas the other systems need a certain minimum base line. In contrast to laser scanning systems, no mechanical moving parts are needed.

Efficient distance algorithm

It is very easy to extract the distance information out of the output signals of the TOF sensor, therefore this task uses only a small amount of processing power, again in contrast to stereo vision, where complex correlation algorithms have to be implemented. After the distance data has been extracted, object detection, for example, is also easy to carry out because the algorithms are not disturbed by patterns on the object.

Speed

Time-of-flight cameras are able to measure the distances within a complete scene with one shot. As the cameras reach up to 100 frames per second, they are ideally suited to be used in real-time applications.

Disadvantages

Background light

Although most of the background light coming from artificial lighting or the sun is suppressed, the pixel still has to provide a high dynamic range. The background light also generates electrons, which have to be stored. For example, the illumination units in today's TOF cameras can provide an illumination level of about 1 watt. The Sun has an illumination power of about 50 watts per square meter after the optical bandpass filter. Therefore, if the illuminated scene has a size of 1 square meter, the light from the sun is 50 times stronger than the modulated signal.

Interference

If several time-of-flight cameras are running at the same time, the cameras may disturb each others' measurements. There exist several possibilities for dealing with this problem:

Multiple reflections

In contrast to laser scanning systems, where only a single point is illuminated at once, the time-of-flight cameras illuminate a whole scene. Due to multiple reflections, the light may reach the objects along several paths and therefore, the measured distance may be greater than the true distance.

Applications

Automotive applications

Time-of-flight cameras are also used in assistance and safety functions for advanced automotive applications such as active pedestrian safety, precrash detection and indoor applications like out-of-position (OOP) detection.[18][19][20]

Human-machine interfaces and gaming

As time-of-flight cameras provide distance images in real time, it is easy to track movements of humans. This allows new interactions with consumer devices such as televisions. Another topic is to use this type of cameras to interact with games on video game consoles.[21]

Measurement and machine vision

Other applications are measurement tasks, e.g. for the fill height in silos. In industrial machine vision, the time-of-flight camera helps to classify objects and help robots find the items, for instance on a conveyor. Door controls can distinguish easily between animals and humans reaching the door.

Robotics

Another use of these cameras is the field of robotics: Mobile robots can build up a map of their surroundings very quickly, enabling them to avoid obstacles or follow a leading person. As the distance calculation is simple, only little computational power is used.

Brands

Active brands (as of 2011)
Defunct brands

See also

References

  1. ^ a b Iddan, Gavriel J.; Yahav, Giora (2001-01-24). "3D imaging in the studio (and elsewhere…)". Proceedings of SPIE (San Jose, CA: SPIE) 4298: pp. 48. 2003-04-29. doi:10.1117/12.424913. Archived from the original on 2009-06-12. http://web.archive.org/web/20090612071500/http://www.3dvsystems.com/technology/3D%20Imaging%20in%20the%20studio.pdf. Retrieved 2009-08-17. "The [time-of-flight] camera belongs to a broader group of sensors known as scanner-less LIDAR (i.e. laser radar having no mechanical scanner); an early [1990] example is [Marion W.] Scott and his followers at Sandia." 
  2. ^ "Product Evolution". 3DV Systems. Archived from the original on 2009-02-28. http://web.archive.org/web/20090228203547/http://www.3dvsystems.com/technology/product.html#1. Retrieved 2009-02-19. "Z-Cam, the first depth video camera, was released in 2000 and was targeted primarily at broadcasting organizations." 
  3. ^ Schuon, Sebastian; Theobalt, Christian; Davis, James; Thrun, Sebastian (2008-07-15). "High-quality scanning using time-of-flight depth superresolution". written at Anchorage, Alaska. IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, 2008. Institute of Electrical and Electronics Engineers. pp. 1–7. doi:10.1109/CVPRW.2008.4563171. ISBN 978-1-4244-2339-2. http://www-cs.stanford.edu/people/theobalt/TOF_CV_Superresolution_final.pdf. Retrieved 2009-07-31. "The Z-cam can measure full frame depth at video rate and at a resolution of 320×240 pixels." 
  4. ^ "Canesta's latest 3D Sensor - "Cobra" ... highest res CMOS 3D depth sensor in the world" (Flash Video). Sunnyvale, California: Canesta. 2010-10-25. http://www.youtube.com/watch?v=5_PVx1NbUZQ. "Canesta "Cobra" 320 x 200 Depth Sensor, capable of 1mm depth resolution, USB powered, 30 to 100 fps […] The complete camera module is about the size of a silver dollar" 
  5. ^ PMD[vision] CamCube 2.0 Datasheet (No. 20090601 ed.). Siegen, Germany: PMDTechnologies. 2009-06-01. p. 5. http://www.pmdtec.com/fileadmin/pmdtec/downloads/documentation/datasheet_camcube.pdf. Retrieved 2009-07-31. "Type of Sensor: PhotonICs PMD 41k-S (204 x 204)" 
  6. ^ SR4000 Data Sheet (Rev 2.6 ed.). Zürich, Switzerland: Mesa Imaging. August 2009. p. 1. http://www.mesa-imaging.ch/dlm.php?fname=pdf/SR4000_Data_Sheet.pdf. Retrieved 2009-08-18. "176 x 144 pixel array (QCIF)" 
  7. ^ PMD[vision] S3 Datasheet (No. 20090601 ed.). Siegen, Germany: PMDTechnologies. 2009-06-01. p. 4. http://www.pmdtec.com/fileadmin/pmdtec/downloads/documentation/datasheet_s3.pdf. Retrieved 2009-07-31. "Type of sensor—PhotonICs 3k-S2, resolution: 64[V] x 48[H] pixels" 
  8. ^ "2+3D - real world in real time". odos imaging. http://www.odos-imaging.com/view.php/page/index. Retrieved 2011-10-20. "With capability of up to 100 frames per second, very fast moving objects can be captured." 
  9. ^ Stettner, Roger and Bailey, Howard. Eye-safe laser radar 3D imaging. Advanced Scientific Concepts. http://www.advancedscientificconcepts.com/technology/documents/Eye-safepaper.pdf. 
  10. ^ Christoph Heckenkamp: Das magische Auge - Grundlagen der Bildverarbeitung: Das PMD Prinzip. In: Inspect. Nr. 1, 2008, S. 25–28.
  11. ^ a b Gokturk, Salih Burak; Yalcin, Hakan; Bamji, Cyrus (2005-01-24). "A Time-Of-Flight Depth Sensor - System Description, Issues and Solutions". IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, 2004. Institute of Electrical and Electronics Engineers. pp. 35–45. doi:10.1109/CVPR.2004.291. Archived from the original on 2007-06-23. http://www.canesta.com/assets/pdf/technicalpapers/CVPR_Submission_TOF.pdf. Retrieved 2009-07-31. "The differential structure accumulates photo-generated charges in two collection nodes using two modulated gates. The gate modulation signals are synchronized with the light source, and hence depending on the phase of incoming light, one node collects more charges than the other. At the end of integration, the voltage difference between the two nodes is read out as a measure of the phase of the reflected light." 
  12. ^ "Mesa Imaging - Products". August 17, 2009. http://www.mesa-imaging.ch. 
  13. ^ Medina A, Gayá F, and Pozo F. Compact laser radar and three-dimensional camera. 23 (2006). J. Opt. Soc. Am. A. pp. 800–805. http://www.opticsinfobase.org/abstract.cfm?URI=josaa-23–4–800. 
  14. ^ Medina, Antonio. Three Dimensional Camera and Rangefinder. January 1992. United States Patent 5081530. 
  15. ^ http://www.laseroptronix.se/gated/sealynx.pdf
  16. ^ "Sea-Lynx Gated Camera - active laser camera system". http://www.laseroptronix.se/gated/sealynx.pdf. 
  17. ^ "CCD/CMOS Lock-In Pixel for Range Imaging: Challenges, Limitations and State-of-the-Art" - CSEM
  18. ^ Hsu, Stephen; Acharya, Sunil; Rafii, Abbas; New, Richard (2006-04-25). "Performance of a Time-of-Flight Range Camera for Intelligent Vehicle Safety Applications". Advanced Microsystems for Automotive Applications 2006. Springer. pp. 205–219. doi:10.1007/3-540-33410-6_16. ISBN 978-3-540-33410-1. Archived from the original on 2006-12-06. http://www.canesta.com/assets/pdf/technicalpapers/canesta_amaa06_paper_final1.pdf. Retrieved 2009-09-36. 
  19. ^ "PMDTec's 3D ToF technology for AUDI: Tomorrow’s driver assistance and safety systems" (Press release). PMDTechnologies. 2011-11-21. http://www.pmdtec.com/news-press/news/detail/pmdtecs-3d-tof-technology-for-audi-tomorrows-driver-assistance-and-safety-systems/. 
  20. ^ Elkhalili, Omar; Schrey, Olaf M.; Ulfig, Wiebke; Brockherde, Werner; Hosticka, Bedrich J. (September 2006), "A 64x8 pixel 3-D CMOS time-of flight image sensor for car safety applications", European Solid State Circuits Conference 2006, pp. 568–571, doi:10.1109/ESSCIR.2006.307488, ISBN 978-1-4244-0302-8, http://publica.fraunhofer.de/documents/N-48683.html, retrieved 2010-03-05 
  21. ^ Captain, Sean (2008-05-01). "Out of Control Gaming". PopSci.com. Popular Science. http://www.popsci.com/gear-gadgets/article/2008-05/out-control-gaming. Retrieved 2009-06-15. 
  22. ^ http://pewa.panasonic.com/components/built-in-sensors/3d-image-sensors/d-imager/
  23. ^ http://www.softkinetic.com/Solutions/DepthSensecameras.aspx
  24. ^ http://www.fotonic.com/content/Products/Default.aspx
  25. ^ http://www.pmdtec.com/
  26. ^ http://www.mesa-imaging.ch/prodview4k.php
  27. ^ http://www.iee.lu/technologies
  28. ^ http://www.odos-imaging.com
  29. ^ http://www.brainvision.co.jp/xoops/modules/tinyd4/index.php?id=5
  30. ^ http://www.tridicam.net/en/products/array-sensor

External links